Uniform Distribution (continuous)
   HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
and
statistics Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
, the continuous uniform distribution or rectangular distribution is a family of
symmetric Symmetry (from grc, συμμετρία "agreement in dimensions, due proportion, arrangement") in everyday language refers to a sense of harmonious and beautiful proportion and balance. In mathematics, "symmetry" has a more precise definiti ...
probability distributions In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon ...
. The distribution describes an experiment where there is an arbitrary outcome that lies between certain bounds. The bounds are defined by the parameters, ''a'' and ''b'', which are the minimum and maximum values. The interval can either be closed (e.g. , b or
open Open or OPEN may refer to: Music * Open (band), Australian pop/rock band * The Open (band), English indie rock band * ''Open'' (Blues Image album), 1969 * ''Open'' (Gotthard album), 1999 * ''Open'' (Cowboy Junkies album), 2001 * ''Open'' (YF ...
(e.g. (a, b)). Therefore, the distribution is often abbreviated ''U'' (''a'', ''b''), where U stands for uniform distribution. The difference between the bounds defines the interval length; all intervals of the same length on the distribution's support are equally probable. It is the
maximum entropy probability distribution In statistics and information theory, a maximum entropy probability distribution has entropy that is at least as great as that of all other members of a specified class of probability distributions. According to the principle of maximum entro ...
for a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
''X'' under no constraint other than that it is contained in the distribution's support.


Definitions


Probability density function

The
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
of the continuous uniform distribution is: : f(x)=\begin \frac & \mathrm\ a \le x \le b, \\ pt 0 & \mathrm\ xb \end The values of ''f''(''x'') at the two boundaries ''a'' and ''b'' are usually unimportant because they do not alter the values of the integrals of over any interval, nor of or any higher moment. Sometimes they are chosen to be zero, and sometimes chosen to be . The latter is appropriate in the context of estimation by the method of
maximum likelihood In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed stat ...
. In the context of Fourier analysis, one may take the value of ''f''(''a'') or ''f''(''b'') to be , since then the inverse transform of many
integral transform In mathematics, an integral transform maps a function from its original function space into another function space via integration, where some of the properties of the original function might be more easily characterized and manipulated than in ...
s of this uniform function will yield back the function itself, rather than a function which is equal "
almost everywhere In measure theory (a branch of mathematical analysis), a property holds almost everywhere if, in a technical sense, the set for which the property holds takes up nearly all possibilities. The notion of "almost everywhere" is a companion notion to ...
", i.e. except on a set of points with zero
measure Measure may refer to: * Measurement, the assignment of a number to a characteristic of an object or event Law * Ballot measure, proposed legislation in the United States * Church of England Measure, legislation of the Church of England * Mea ...
. Also, it is consistent with the sign function which has no such ambiguity. Graphically, the
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
is portrayed as a rectangle where is the base and is the height. As the distance between a and b increases, the density at any particular value within the distribution boundaries decreases. Since the
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
integrates to 1, the height of the probability density function decreases as the base length increases. In terms of mean ''μ'' and variance ''σ''2, the probability density may be written as: : f(x)=\begin \frac & \mbox-\sigma\sqrt \le x-\mu \le \sigma\sqrt \\ 0 & \text \end


Cumulative distribution function

The
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
is: : F(x)= \begin 0 & \textx < a \\ pt \frac & \texta \le x \le b \\ pt 1 & \textx > b \end Its inverse is: :F^(p) = a + p (b - a) \,\,\text 0 In mean and variance notation, the cumulative distribution function is: :F(x)= \begin 0 & \textx-\mu < -\sigma\sqrt \\ \frac \left( \frac +1 \right) & \text-\sigma\sqrt \le x-\mu < \sigma\sqrt \\ 1 & \textx-\mu \ge \sigma\sqrt \end and the inverse is: :F^(p) = \sigma\sqrt(2p-1) +\mu\,\, \text0 \le p \le 1


Example 1. Using the Uniform Cumulative Distribution Function

For random variable : X\sim U(0,23) Find \scriptstyle P(2 < X < 18): : P(2 < X < 18) = (18-2)\cdot \frac 1 = \frac . In graphical representation of uniform distribution function (x) vs x the area under the curve within the specified bounds displays the probability (shaded area is depicted as a rectangle). For this specific example above, the base would be and the height would be .


Example 2. Using the Uniform Cumulative Distribution Function (Conditional)

For random variable : X\sim U(0,23) Find \scriptstyle P(X > 12 \ , \ X > 8): : P(X > 12\ , \ X > 8) = (23-12)\cdot \frac 1 = \frac. The example above is for a conditional probability case for the uniform distribution: given is true, what is the probability that . Conditional probability changes the sample space so a new interval length has to be calculated, where is 23 and is 8. The graphical representation would still follow Example 1, where the area under the curve within the specified bounds displays the probability and the base of the rectangle would be and the height .


Generating functions


Moment-generating function

The
moment-generating function In probability theory and statistics, the moment-generating function of a real-valued random variable is an alternative specification of its probability distribution. Thus, it provides the basis of an alternative route to analytical results compare ...
is: : M_x = E(e^) = \frac \,\! from which we may calculate the
raw moments In mathematics, the moments of a function are certain quantitative measures related to the shape of the function's graph. If the function represents mass density, then the zeroth moment is the total mass, the first moment (normalized by total ma ...
''m'' ''k'' :m_1=\frac, \,\! :m_2=\frac, \,\! :m_k=\frac\sum_^k a^ib^. \,\! For the special case ''a'' = –''b'', that is, for : f(x)=\begin \frac & \text\ -b \le x \le b, \\ pt 0 & \text, \end the moment-generating functions reduces to the simple form : M_x=\frac. For a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
following this distribution, the
expected value In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a l ...
is then ''m''1 = (''a'' + ''b'')/2 and the
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbe ...
is ''m''2 − ''m''12 = (''b'' − ''a'')2/12.


Cumulant-generating function

For , the ''n''th
cumulant In probability theory and statistics, the cumulants of a probability distribution are a set of quantities that provide an alternative to the '' moments'' of the distribution. Any two probability distributions whose moments are identical will have ...
of the uniform distribution on the interval is ''B''''n''/''n'', where ''B''''n'' is the ''n''th
Bernoulli number In mathematics, the Bernoulli numbers are a sequence of rational numbers which occur frequently in analysis. The Bernoulli numbers appear in (and can be defined by) the Taylor series expansions of the tangent and hyperbolic tangent functions, ...
.


Standard uniform

Restricting a=0 and b=1, the resulting distribution ''U''(0,1) is called a standard uniform distribution. One interesting property of the standard uniform distribution is that if ''u''1 has a standard uniform distribution, then so does 1-''u''1. This property can be used for generating
antithetic variates In statistics, the antithetic variates method is a variance reduction technique used in Monte Carlo methods. Considering that the error in the simulated signal (using Monte Carlo methods) has a one-over square root convergence, a very large number ...
, among other things. In other words, this property is known as the
inversion method Inverse transform sampling (also known as inversion sampling, the inverse probability integral transform, the inverse transformation method, Smirnov transform, or the golden ruleAalto University, N. Hyvönen, Computational methods in inverse probl ...
where the continuous standard uniform distribution can be used to generate random numbers for any other continuous distribution. If is a uniform random number with standard uniform distribution (0,1), then x= F^(u) generates a random number from any continuous distribution with the specified
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
.


Relationship to other functions

As long as the same conventions are followed at the transition points, the probability density function may also be expressed in terms of the
Heaviside step function The Heaviside step function, or the unit step function, usually denoted by or (but sometimes , or ), is a step function, named after Oliver Heaviside (1850–1925), the value of which is zero for negative arguments and one for positive argum ...
: :f(x)=\frac, \,\! or in terms of the
rectangle function The rectangular function (also known as the rectangle function, rect function, Pi function, Heaviside Pi function, gate function, unit pulse, or the normalized boxcar function) is defined as \operatorname(t) = \Pi(t) = \left\{\begin{array}{r ...
:f(x)=\frac\,\operatorname\left(\frac\right) . There is no ambiguity at the transition point of the sign function. Using the half-maximum convention at the transition points, the uniform distribution may be expressed in terms of the sign function as: :f(x)=\frac .


Properties


Moments

The mean (first moment) of the distribution is: :E(X)=\frac(b+a). The second moment of the distribution is: :E(X^2) = \frac. In general, the ''n''-th moment of the uniform distribution is: :E(X^n) = \frac The variance (second
central moment In probability theory and statistics, a central moment is a moment of a probability distribution of a random variable about the random variable's mean; that is, it is the expected value of a specified integer power of the deviation of the random ...
) is: :V(X)=\frac(b-a)^2


Order statistics

Let ''X''1, ..., ''X''''n'' be an
i.i.d. In probability theory and statistics, a collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent. This property is us ...
sample from ''U''(0,1). Let ''X''(''k'') be the ''k''th
order statistic In statistics, the ''k''th order statistic of a statistical sample is equal to its ''k''th-smallest value. Together with rank statistics, order statistics are among the most fundamental tools in non-parametric statistics and inference. Importan ...
from this sample. Then the probability distribution of ''X''(''k'') is a
Beta distribution In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval , 1in terms of two positive parameters, denoted by ''alpha'' (''α'') and ''beta'' (''β''), that appear as ...
with parameters ''k'' and . The expected value is :\operatorname(X_) = . This fact is useful when making
Q–Q plot In statistics, a Q–Q plot (quantile-quantile plot) is a probability plot, a graphical method for comparing two probability distributions by plotting their ''quantiles'' against each other. A point on the plot corresponds to one of the qu ...
s. The variances are :\operatorname(X_) = . See also:


Uniformity

The probability that a uniformly distributed random variable falls within any interval of fixed length is independent of the location of the interval itself (but it is dependent on the interval size), so long as the interval is contained in the distribution's support. To see this, if ''X'' ~ U(''a'',''b'') and 'x'', ''x''+''d''is a subinterval of 'a'',''b''with fixed ''d'' > 0, then : P\left(X\in\left x,x+d \right right) = \int_^ \frac\, = \frac \,\! which is independent of ''x''. This fact motivates the distribution's name.


Generalization to Borel sets

This distribution can be generalized to more complicated sets than intervals. If ''S'' is a
Borel set In mathematics, a Borel set is any set in a topological space that can be formed from open sets (or, equivalently, from closed sets) through the operations of countable union, countable intersection, and relative complement. Borel sets are named ...
of positive, finite measure, the uniform probability distribution on ''S'' can be specified by defining the pdf to be zero outside ''S'' and constantly equal to 1/''K'' on ''S'', where ''K'' is the
Lebesgue measure In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of ''n''-dimensional Euclidean space. For ''n'' = 1, 2, or 3, it coincides wit ...
of ''S''.


Related distributions

* If ''X'' has a standard uniform distribution, then by the
inverse transform sampling Inverse transform sampling (also known as inversion sampling, the inverse probability integral transform, the inverse transformation method, Smirnov transform, or the golden ruleAalto University, N. Hyvönen, Computational methods in inverse probl ...
method, ''Y'' = − λ−1 ln(X) has an
exponential distribution In probability theory and statistics, the exponential distribution is the probability distribution of the time between events in a Poisson point process, i.e., a process in which events occur continuously and independently at a constant average ...
with (rate) parameter λ. * If ''X'' has a standard uniform distribution, then ''Y'' = ''X''''n'' has a
beta distribution In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval , 1in terms of two positive parameters, denoted by ''alpha'' (''α'') and ''beta'' (''β''), that appear as ...
with parameters (''1/n,1)''. As such, * The standard uniform distribution is a special case of the
beta distribution In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval , 1in terms of two positive parameters, denoted by ''alpha'' (''α'') and ''beta'' (''β''), that appear as ...
with parameters (''1,1)''. * The
Irwin–Hall distribution In probability and statistics, the Irwin–Hall distribution, named after Joseph Oscar Irwin and Philip Hall, is a probability distribution for a random variable defined as the sum of a number of independent random variables, each having a unifo ...
is the sum of ''n''
i.i.d. In probability theory and statistics, a collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent. This property is us ...
''U(0,1)'' distributions. * The sum of two independent, equally distributed, uniform distributions yields a symmetric
triangular distribution In probability theory and statistics, the triangular distribution is a continuous probability distribution with lower limit ''a'', upper limit ''b'' and mode ''c'', where ''a'' < ''b'' and ''a'' ≤ ''c'' ≤ ''b''. ...
. * The distance between two
i.i.d. In probability theory and statistics, a collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent. This property is us ...
uniform random variables also has a
triangular distribution In probability theory and statistics, the triangular distribution is a continuous probability distribution with lower limit ''a'', upper limit ''b'' and mode ''c'', where ''a'' < ''b'' and ''a'' ≤ ''c'' ≤ ''b''. ...
, although not symmetric.


Statistical inference


Estimation of parameters


Estimation of maximum


=Minimum-variance unbiased estimator

= Given a uniform distribution on , ''b''with unknown ''b,'' the minimum-variance unbiased estimator (UMVUE) for the maximum is given by :\hat_\text=\frac m = m + \frac where ''m'' is the
sample maximum In statistics, the sample maximum and sample minimum, also called the largest observation and smallest observation, are the values of the greatest and least elements of a sample. They are basic summary statistics, used in descriptive statistic ...
and ''k'' is the
sample size Sample size determination is the act of choosing the number of observations or Replication (statistics), replicates to include in a statistical sample. The sample size is an important feature of any empirical study in which the goal is to make stat ...
, sampling without replacement (though this distinction almost surely makes no difference for a continuous distribution). This follows for the same reasons as estimation for the discrete distribution, and can be seen as a very simple case of
maximum spacing estimation In statistics, maximum spacing estimation (MSE or MSP), or maximum product of spacing estimation (MPS), is a method for estimating the parameters of a univariate parametric model, statistical model. The method requires maximization of the geometr ...
. This problem is commonly known as the
German tank problem In the statistical theory of estimation theory, estimation, the German tank problem consists of estimating the maximum of a discrete uniform distribution from sampling without replacement. In simple terms, suppose there exists an unknown number o ...
, due to application of maximum estimation to estimates of German tank production during
World War II World War II or the Second World War, often abbreviated as WWII or WW2, was a world war that lasted from 1939 to 1945. It involved the vast majority of the world's countries—including all of the great powers—forming two opposin ...
.


=Maximum likelihood estimator

= The
maximum likelihood In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed stat ...
estimator is given by: :\hat_= m where ''m'' is the
sample maximum In statistics, the sample maximum and sample minimum, also called the largest observation and smallest observation, are the values of the greatest and least elements of a sample. They are basic summary statistics, used in descriptive statistic ...
, also denoted as m=X_ the maximum
order statistic In statistics, the ''k''th order statistic of a statistical sample is equal to its ''k''th-smallest value. Together with rank statistics, order statistics are among the most fundamental tools in non-parametric statistics and inference. Importan ...
of the sample.


=Method of moment estimator

= The method of moments estimator is given by: :\hat_= 2\bar where \bar is the sample mean.


Estimation of midpoint

The midpoint of the distribution (''a'' + ''b'') / 2 is both the mean and the median of the uniform distribution. Although both the sample mean and the sample median are
unbiased estimator In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
s of the midpoint, neither is as efficient as the sample
mid-range In statistics, the mid-range or mid-extreme is a measure of central tendency of a sample defined as the arithmetic mean of the maximum and minimum values of the data set: :M=\frac. The mid-range is closely related to the range, a measure of ...
, i.e. the arithmetic mean of the sample maximum and the sample minimum, which is the UMVU estimator of the midpoint (and also the
maximum likelihood estimate In statistics, maximum likelihood estimation (MLE) is a method of estimating the parameters of an assumed probability distribution, given some observed data. This is achieved by maximizing a likelihood function so that, under the assumed statist ...
).


Confidence interval


For the maximum

Let ''X''1, ''X''2, ''X''3, ..., ''X''''n'' be a sample from U_ where ''L'' is the population maximum. Then ''X''(''n'') = max( ''X''1, ''X''2, ''X''3, ..., ''X''''n'' ) has the Lebesgue-Borel-density f := \frac Nechval KN, Nechval NA, Vasermanis EK, Makeev VY (2002
Constructing shortest-length confidence intervals
Transport and Telecommunication 3 (1) 95-103
: f(t) = n \frac \left(\frac\right)^ =n \frac, 0 \leq t \leq L The confidence interval given before is mathematically incorrect, as \Pr ( hat, \hat + \epsilon\ni \theta) \geq 1 - \alpha cannot be solved for \epsilon without knowledge of \theta. However one can solve : \Pr ( hat, \hat (1 + \epsilon)\ni \theta) \geq 1 - \alpha for \epsilon \geq (1-\alpha)^ - 1 for any unknown but valid \theta, one then chooses the smallest \epsilon possible satisfying the condition above. Note that the interval length depends upon the random variable \hat.


Occurrence and applications

The probabilities for uniform distribution function are simple to calculate due to the simplicity of the function form. Therefore, there are various applications that this distribution can be used for as shown below: hypothesis testing situations, random sampling cases, finance, etc. Furthermore, generally, experiments of physical origin follow a uniform distribution (e.g. emission of radioactive
particles In the physical sciences, a particle (or corpuscule in older texts) is a small localized object which can be described by several physical or chemical properties, such as volume, density, or mass. They vary greatly in size or quantity, from s ...
). However, it is important to note that in any application, there is the unchanging assumption that the probability of falling in an interval of fixed length is constant.


Economics example for uniform distribution

In the field of economics, usually
demand In economics, demand is the quantity of a good that consumers are willing and able to purchase at various prices during a given time. The relationship between price and quantity demand is also called the demand curve. Demand for a specific item ...
and replenishment may not follow the expected normal distribution. As a result, other distribution models are used to better predict probabilities and trends such as Bernoulli process. But according to Wanke (2008), in the particular case of investigating lead-time for inventory management at the beginning of the
life cycle Life cycle, life-cycle, or lifecycle may refer to: Science and academia *Biological life cycle, the sequence of life stages that an organism undergoes from birth to reproduction ending with the production of the offspring * Life-cycle hypothesis ...
when a completely new product is being analyzed, the uniform distribution proves to be more useful. In this situation, other distribution may not be viable since there is no existing data on the new product or that the demand history is unavailable so there isn't really an appropriate or known distribution. The uniform distribution would be ideal in this situation since the random variable of lead-time (related to demand) is unknown for the new product but the results are likely to range between a plausible range of two values. The lead-time would thus represent the random variable. From the uniform distribution model, other factors related to lead-time were able to be calculated such as cycle service level and shortage per cycle. It was also noted that the uniform distribution was also used due to the simplicity of the calculations.


Sampling from an arbitrary distribution

The uniform distribution is useful for sampling from arbitrary distributions. A general method is the inverse transform sampling method, which uses the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
(CDF) of the target random variable. This method is very useful in theoretical work. Since simulations using this method require inverting the CDF of the target variable, alternative methods have been devised for the cases where the CDF is not known in closed form. One such method is
rejection sampling In numerical analysis and computational statistics, rejection sampling is a basic technique used to generate observations from a distribution. It is also commonly called the acceptance-rejection method or "accept-reject algorithm" and is a type of ...
. The
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
is an important example where the inverse transform method is not efficient. However, there is an exact method, the Box–Muller transformation, which uses the inverse transform to convert two independent uniform
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s into two independent normally distributed random variables.


Quantization error

In analog-to-digital conversion a quantization error occurs. This error is either due to rounding or truncation. When the original signal is much larger than one least significant bit (LSB), the quantization error is not significantly correlated with the signal, and has an approximately uniform distribution. The RMS error therefore follows from the variance of this distribution.


Random variate generation

There are many applications in which it is useful to run simulation experiments. Many
programming language A programming language is a system of notation for writing computer programs. Most programming languages are text-based formal languages, but they may also be graphical. They are a kind of computer language. The description of a programming ...
s come with implementations to generate
pseudo-random numbers A pseudorandom sequence of numbers is one that appears to be statistically random, despite having been produced by a completely deterministic and repeatable process. Background The generation of random numbers has many uses, such as for random ...
which are effectively distributed according to the standard uniform distribution. On the other hand, the uniformly distributed numbers are often used as the basis for
non-uniform random variate generation Non-uniform random variate generation or pseudo-random number sampling is the numerical practice of generating pseudo-random numbers (PRN) that follow a given probability distribution. Methods are typically based on the availability of a unifor ...
. If ''u'' is a value sampled from the standard uniform distribution, then the value ''a'' + (''b'' − ''a'')''u'' follows the uniform distribution parametrised by ''a'' and ''b'', as described above.


History

While the historical origins in the conception of uniform distribution are inconclusive, it is speculated that the term 'uniform' arose from the concept of
equiprobability Equiprobability is a property for a collection of events that each have the same probability of occurring. In statistics and probability theory it is applied in the discrete uniform distribution and the equidistribution theorem for rational nu ...
in dice games (note that the dice games would have
discrete Discrete may refer to: *Discrete particle or quantum in physics, for example in quantum theory * Discrete device, an electronic component with just one circuit element, either passive or active, other than an integrated circuit *Discrete group, a ...
and not continuous uniform sample space).
Equiprobability Equiprobability is a property for a collection of events that each have the same probability of occurring. In statistics and probability theory it is applied in the discrete uniform distribution and the equidistribution theorem for rational nu ...
was mentioned in Gerolamo Cardano's ''Liber de Ludo Aleae'', a manual written in 16th century and detailed on advanced probability calculus in relation to dice.


See also

* Discrete uniform distribution *
Beta distribution In probability theory and statistics, the beta distribution is a family of continuous probability distributions defined on the interval , 1in terms of two positive parameters, denoted by ''alpha'' (''α'') and ''beta'' (''β''), that appear as ...
*
Box–Muller transform The Box–Muller transform, by George Edward Pelham Box and Mervin Edgar Muller, is a random number sampling method for generating pairs of independent, standard, normally distributed (zero expectation, unit variance) random numbers, given a ...
* Probability plot *
Q–Q plot In statistics, a Q–Q plot (quantile-quantile plot) is a probability plot, a graphical method for comparing two probability distributions by plotting their ''quantiles'' against each other. A point on the plot corresponds to one of the qu ...
*
Rectangular function The rectangular function (also known as the rectangle function, rect function, Pi function, Heaviside Pi function, gate function, unit pulse, or the normalized boxcar function) is defined as \operatorname(t) = \Pi(t) = \left\{\begin{array}{r ...
*
Irwin–Hall distribution In probability and statistics, the Irwin–Hall distribution, named after Joseph Oscar Irwin and Philip Hall, is a probability distribution for a random variable defined as the sum of a number of independent random variables, each having a unifo ...
— In the degenerate case where n=1, the Irwin-Hall distribution generates a uniform distribution between 0 and 1. *
Bates distribution In probability and business statistics, the Bates distribution, named after Grace Bates, is a probability distribution of the mean of a number of statistically independent uniformly distributed random variables on the unit interval. This dist ...
— Similar to the Irwin-Hall distribution, but rescaled for n. Like the Irwin-Hall distribution, in the degenerate case where n=1, the Bates distribution generates a uniform distribution between 0 and 1.


References


Further reading

*


External links


Online calculator of Uniform distribution (continuous)
{{DEFAULTSORT:Uniform Distribution (Continuous) Continuous distributions Location-scale family probability distributions su:Sebaran seragam#Kasus kontinyu